Reinforcement Learning-Based Sequential Batch-Sampling for Bayesian Optimal Experimental Design
نویسندگان
چکیده
Abstract Engineering problems that are modeled using sophisticated mathematical methods or characterized by expensive-to-conduct tests experiments encumbered with limited budget finite computational resources. Moreover, practical scenarios in the industry, impose restrictions, based on logistics and preference, manner which can be conducted. For example, material supply may enable only a handful of single-shot case models one face significant wait-time shared In such scenarios, usually resorts to performing allows for maximizing one’s state-of-knowledge while satisfying above-mentioned constraints. Sequential design (SDOE) is popular suite have yielded promising results recent years across different engineering problems. A common strategy leverages Bayesian formalism SDOE, works best one-step-ahead myopic scenario selecting single experiment at each step sequence experiments. this work, we aim extend SDOE strategy, query computer code batch inputs. To end, leverage deep reinforcement learning (RL)-based policy gradient methods, propose batches queries selected taking into account entire hand. The algorithm retains sequential nature, inherent incorporating elements reward task from domain RL. unique capability proposed methodology its ability applied multiple tasks, optimization function, once trained. We demonstrate performance synthetic problem challenging high-dimensional problem.
منابع مشابه
Numerical Approaches for Sequential Bayesian Optimal Experimental Design
Experimental data play a crucial role in developing and refining models of physical systems. Some experiments can be more valuable than others, however. Well-chosen experiments can save substantial resources, and hence optimal experimental design (OED) seeks to quantify and maximize the value of experimental data. Common current practice for designing a sequence of experiments uses suboptimal a...
متن کاملSequential Bayesian optimal experimental design via approximate dynamic programming
The design of multiple experiments is commonly undertaken via suboptimal strategies, such as batch (open-loop) design that omits feedback or greedy (myopic) design that does not account for future effects. This paper introduces new strategies for the optimal design of sequential experiments. First, we rigorously formulate the general sequential optimal experimental design (sOED) problem as a dy...
متن کاملSmarter Sampling in Model-Based Bayesian Reinforcement Learning
Bayesian reinforcement learning (RL) is aimed at making more efficient use of data samples, but typically uses significantly more computation. For discrete Markov Decision Processes, a typical approach to Bayesian RL is to sample a set of models from an underlying distribution, and compute value functions for each, e.g. using dynamic programming. This makes the computation cost per sampled mode...
متن کاملOptimal Sample Selection for Batch-mode Reinforcement Learning
We introduce the Optimal Sample Selection (OSS) meta-algorithm for solving discrete-time Optimal Control problems. This meta-algorithm maps the problem of finding a near-optimal closed-loop policy to the identification of a small set of one-step system transitions, leading to high-quality policies when used as input of a batch-mode Reinforcement Learning (RL) algorithm. We detail a particular i...
متن کاملSimulation-based optimal Bayesian experimental design for nonlinear systems
The optimal selection of experimental conditions is essential to maximizing the value of data for inference and prediction, particularly in situations where experiments are time-consuming and expensive to conduct. We propose a general mathematical framework and an algorithmic approach for optimal experimental design with nonlinear simulation-based models; in particular, we focus on finding sets...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Journal of Mechanical Design
سال: 2022
ISSN: ['1528-9001', '1050-0472']
DOI: https://doi.org/10.1115/1.4054631